# Continuous Pretraining Optimization
Finance Llama3 8B
A financial domain model developed based on Llama3-8B, enhanced with an instruction pretraining framework to improve domain adaptability, achieving or even surpassing the performance of Llama3-70B on financial tasks.
Large Language Model
Transformers English

F
instruction-pretrain
1,200
64
Layerskip Llama2 7B
Other
An improved model based on Llama2 7B, supporting hierarchical skip and self-speculative decoding to enhance inference efficiency
Large Language Model
Transformers English

L
facebook
1,674
14
Quietstar 8 Ahead
Based on the Mistral-7b model, it employs the Quiet-STaR method for continuous pretraining, generating 8 reasoning tokens before each output token to enhance reasoning capabilities.
Large Language Model
Transformers

Q
ezelikman
239
90
Featured Recommended AI Models